Improving Lasso for model selection and prediction
نویسندگان
چکیده
It is known that the Thresholded Lasso (TL), SCAD or MCP correct intrinsic estimation bias of Lasso. In this paper we propose an alternative method improving for predictive models with general convex loss functions which encompass normal linear models, logistic regression, quantile support vector machines. For a given penalty order absolute values nonzero coefficients and then select final model from small nested family by Generalized Information Criterion. We derive exponential upper bounds on selection error method. These results confirm that, at least our algorithm seems to be benchmark theory as it constructive, computationally efficient leads consistent under weak assumptions. Constructivity means in contrast TL, MCP, does not rely unknown parameters cone invertibility factor. Instead, only needs sample size, number predictors bound noise parameter. show numerical experiments synthetic real-world datasets implementation more accurate than implementations studied concave regularizations. Our procedure included R package DMRnet available CRAN repository.
منابع مشابه
On Model Selection Consistency of Lasso On Model Selection Consistency of Lasso
Sparsity or parsimony of statistical models is crucial for their proper interpretations, as in sciences and social sciences. Model selection is a commonly used method to find such models, but usually involves a computationally heavy combinatorial search. Lasso (Tibshirani, 1996) is now being used as a computationally feasible alternative to model selection. Therefore it is important to study La...
متن کاملOn Model Selection Consistency of Lasso
Sparsity or parsimony of statistical models is crucial for their proper interpretations, as in sciences and social sciences. Model selection is a commonly used method to find such models, but usually involves a computationally heavy combinatorial search. Lasso (Tibshirani, 1996) is now being used as a computationally feasible alternative to model selection. Therefore it is important to study La...
متن کاملA Note on the Lasso for Gaussian Graphical Model Selection
Inspired by the success of the Lasso for regression analysis (Tibshirani, 1996), it seems attractive to estimate the graph of a multivariate normal distribution by `1-norm penalised likelihood maximisation. The objective function is convex and the graph estimator can thus be computed efficiently, even for very large graphs. However, we show in this note that the resulting estimator is not consi...
متن کاملImproved Lasso for genomic selection.
Empirical experience with genomic selection in dairy cattle suggests that the distribution of the effects of single nucleotide polymorphisms (SNPs) might be far from normality for some traits. An alternative, avoiding the use of arbitrary prior information, is the Bayesian Lasso (BL). Regular BL uses a common variance parameter for residual and SNP effects (BL1Var). We propose here a BL with di...
متن کاملThe lasso method for variable selection in the Cox model.
I propose a new method for variable selection and shrinkage in Cox's proportional hazards model. My proposal minimizes the log partial likelihood subject to the sum of the absolute values of the parameters being bounded by a constant. Because of the nature of this constraint, it shrinks coefficients and produces some coefficients that are exactly zero. As a result it reduces the estimation vari...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Scandinavian Journal of Statistics
سال: 2021
ISSN: ['0303-6898', '1467-9469']
DOI: https://doi.org/10.1111/sjos.12546